Flexible Priors for Infinite Mixture Models
نویسنده
چکیده
Most infinite mixture models in the current literature are based on the Dirichlet process prior. This prior on partitions implies a very specific (a priori) distribution on cluster sizes. A slightly more general prior known as the Pitman-Yor process prior generalizes this to a two-parameter family. The latter is the most general exchangeable partition probability function (EPPF) as defined by Pitman (Pitman, 2002) known to date. I want to argue that it is desirable to have more flexibility in expressing our prior beliefs over cluster sizes. EPPFs as defined by Pitman satisfy 3 conditions, exchangeability over objects, exchangeability over cluster-labels and consistency. In this contribution I explore the possibility to relax some of these conditions. In particular, I will discuss the consequences of relaxing exchangeability over cluster-labels and consistency. In both cases it turns out that one can formulate proper and efficient Gibbs sampling algorithms but with the added flexibility of having more control to design one’s prior.
منابع مشابه
Slice sampling mixture models
We propose a more efficient version of the slice sampler for Dirichlet process mixture models described by Walker (Commun. Stat., Simul. Comput. 36:45–54, 2007). This new sampler allows for the fitting of infinite mixture models with a wide-range of prior specifications. To illustrate this flexibility we consider priors defined through infinite sequences of independent positive random variables...
متن کاملSome Further Developments for Stick-breaking Priors: Finite and Infinite Clustering and Classification
SUMMARY. The class of stick-breaking priors and their extensions are considered in classification and clustering problems in which the complexity, the number of possible models or clusters, can be either bounded or unbounded. A conjugacy property for the ex tended stick-breaking prior is established which allows for informative characterizations of the priors under i.i.d. sampling, and which fu...
متن کاملAccelerated Variational Infinite Mixture Models
Infinite mixture models, such as the Dirichlet process mixture, are promising candidates for clustering applications where the number of clusters is unknown a priori. Due to computational considerations these models are unfortunately unsuitable for large scale data-mining applications. We propose a class of deterministic accelerated infinite mixture models that can routinely handle millions of ...
متن کاملPosterior Simulation in Countable Mixture Models for Large Datasets
Mixture models, or convex combinations of a countable number of probability distributions, offer an elegant framework for inference when the population of interest can be subdivided into latent clusters having random characteristics that are heterogeneous between, but homogeneous within, the clusters. Traditionally, the different kinds of mixture models have been motivated and analyzed from ver...
متن کاملDefault priors for density estimation with mixture models
The infinite mixture of normals model has become a popular method for density estimation problems. This paper proposes an alternative hierarchical model that leads to hyperparameters that can be interpreted as the location, scale and smoothness of the density. The priors on other parts of the model have little effect on the density estimates and can be given default choices. Automatic Bayesian ...
متن کاملSpike-and-Slab Dirichlet Process Mixture Models
In this paper, Spike-and-Slab Dirichlet Process (SS-DP) priors are introduced and discussed for non-parametric Bayesian modeling and inference, especially in the mixture models context. Specifying a spike-and-slab base measure for DP priors combines the merits of Dirichlet process and spike-and-slab priors and serves as a flexible approach in Bayesian model selection and averaging. Computationa...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2006